Accessibility settings

Published on in Vol 28 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/91940, first published .
Governing Patient-Facing AI-Generated Video in Digital Health: A Risk-and-Ethics Matrix for Deployment, Monitoring, and Change Control

Governing Patient-Facing AI-Generated Video in Digital Health: A Risk-and-Ethics Matrix for Deployment, Monitoring, and Change Control

Governing Patient-Facing AI-Generated Video in Digital Health: A Risk-and-Ethics Matrix for Deployment, Monitoring, and Change Control

Authors of this article:

Yongzheng Hu1, 2 Author Orcid Image ;   Wei Jiang1, 2 Author Orcid Image

1Department of Nephrology, Affiliated Hospital of Qingdao University, 16 Jiangsu Road, Shinan District, Qingdao, Shandong, China

2Department of Nephrology, Qingdao University, Qingdao, Shandong, China

Corresponding Author:

Wei Jiang, MD


In this Viewpoint, we argue that patient-facing high-fidelity artificial intelligence (AI)–generated video requires governance that is operational, life cycle based, and embedded in existing institutional review pathways rather than limited to predeployment checks alone. Patient-facing high-fidelity AI-generated video—synthetic or substantially AI-mediated video that presents realistic human likeness, voices, or clinical communication cues—is rapidly entering patient education and clinical communication. We propose a risk-and-ethics matrix that combines residual clinical risk (likelihood × severity after mitigations) with an ethical alignment score that operationalizes autonomy, beneficence, nonmaleficence, and justice to yield actionable dispositions (encourage, permit with oversight, restrict or redesign, or prohibit). The framework links each disposition to dossier-based review, minimum controls, and postdeployment monitoring triggers—focused on measurable outcomes (eg, comprehension, content-attributable follow-up burden, incidents and complaints, and equity gaps) as well as provenance and change control—to support auditable, revisitable decisions over the system life cycle.

J Med Internet Res 2026;28:e91940

doi:10.2196/91940

Keywords



High-fidelity artificial intelligence (AI)–generated video—from text-to-video patient explainers to deepfake-style clinician avatars—is entering digital health via patient portals, telehealth workflows, and social platforms [1,2]. In this Viewpoint, we use this term to refer to synthetic or substantially AI-mediated video that presents realistic human likeness, voice, or other clinically salient communication cues in ways that may influence patient trust, comprehension, or decisions. We use “operational governance” to mean the institutional processes through which such systems are reviewed, approved, monitored, and re-evaluated over time. “Real-world deployment” refers to routine use outside controlled testing environments, including use through patient portals, telehealth workflows, apps, and social media channels where content may be redistributed or consumed without direct clinician mediation. “Iterative system change” refers to postdeployment modifications to models, prompts, templates, scripts, rendering pipelines, distribution channels, or disclosure and provenance controls that may materially alter system behavior. Early published examples suggest potential value for patient education and communication, including usability-tested patient digital twins for critical care education, avatar-based educational interventions associated with improved parental knowledge and care skills in hydrocephalus, and pilot or specialty use cases in radiology and postoperative patient communication [3-6]. However, governance often lags behind routine deployment, where content is redistributed across channels and iteratively updated (model, prompt, or template changes). These deployments are judged by outcomes beyond technical accuracy. Real-world clinical performance should be assessed using measurable end points that capture patient and system impact, including comprehension (eg, teach back checks), content-attributable follow-up burden, incident and complaint rates, and equity gaps across language and health literacy groups. This gap motivates a workflow-integrated approach that links upfront review to postdeployment monitoring, incident response, and change control across the life cycle [7,8].

The same properties that make AI-generated video attractive in digital health—realism, personalization, and rapid iteration—also create failure modes that are difficult to detect and manage once content is deployed across heterogeneous channels [9,10]. In routine settings, videos may be reposted or clipped outside institutional portals, updated as models and prompt templates change, and consumed without clinician context—conditions that allow for small errors to propagate into clinically consequential misinformation, unnecessary follow-up burden, or delayed care [11-13]. Identity cues embedded in video (eg, clinician likeness, institutional branding, or emotionally resonant avatars) can amplify perceived authority and trust, increasing the impact of misstatements, undisclosed synthetic identity, and privacy misuse. Before introducing residual clinical risk, it is important to distinguish it from clinical risk more broadly. In this paper, “clinical risk” refers to the possibility that patient-facing AI-generated video contributes to clinically relevant harm, including misinformation that changes care, delayed help seeking, unnecessary follow-up burden, privacy or identity misuse, psychological distress, or inequitable performance across patient groups. Risk becomes residual when these foreseeable failure modes are reassessed after proposed safeguards—such as clinician script review, constrained generation, authenticated distribution, disclosure, provenance controls, and escalation pathways—have been specified. Therefore, the matrix classifies the clinically relevant risk that remains after mitigation rather than the unmitigated theoretical hazard.

We propose a workflow-integrated governance approach: the risk-and-ethics matrix. It links residual clinical risk—defined here as the remaining likelihood and severity of clinically relevant harm after mitigation—to a principlism-based ethical alignment score (EAS) to support deployment decisions for patient-facing AI-generated video. Plotting these dimensions yields actionable dispositions (encourage, permit with oversight, restrict or redesign, and prohibit) and links each to minimum controls—dossier-based documentation, disclosure requirements, and human oversight where appropriate—as well as predefined postdeployment monitoring metrics and rereview triggers. We use representative scenarios to show how health systems can translate ethical commitments and probabilistic harms into auditable, revisitable decisions across the life cycle, particularly as content is updated and redistributed beyond its original workflow. We then summarize key risk mechanisms, present the scoring rubric and workflow, and map representative use cases of monitoring and change control actions.

The framework is intended for institutional decision-makers rather than for platform-wide moderation. In this paper, the relevant videos are those created, commissioned, adapted, or sponsored for patient-facing use. Typical producers include health systems; clinicians; patient education teams; digital health vendors; and researchers working in care delivery, education, or protocolized specialist settings. The primary governing bodies are local institutional actors such as institutional review boards (IRBs), digital health or clinical governance committees, patient education and communications leaders, and safety and IT oversight teams. Their role is to decide whether a proposed use case should be approved, under what minimum controls, and with what monitoring and rereview conditions. Existing laws, policies, and AI governance frameworks remain essential, but they often operate at a higher level of abstraction and do not specify how institutions should translate transparency, consent, safety, equity, provenance, and change control expectations into case-level deployment decisions for patient-facing AI-generated video use. What the matrix adds is not a replacement for law or formal regulation but an operational layer for institutional decision-making. It converts broad requirements such as transparency, human oversight, safety, equity, provenance, and accountability into case-level classifications, minimum controls, and life cycle management actions for specific patient-facing AI-generated video use cases [14].

The aim of this Viewpoint is to argue that governance of patient-facing AI-generated video should connect residual risk assessment and ethical alignment to concrete institutional decisions, life cycle monitoring, and change control. To support this argument, we outline the main risk mechanisms; present the risk-and-ethics matrix and its workflow; and then discuss implementation, validation, and adaptation across institutional and regulatory contexts.


When AI-generated video is deployed routinely across patient portals, telehealth workflows, and social platforms, failure modes emerge that are not well captured by predeployment validation and, therefore, require postdeployment monitoring, incident response, and change control. First, misinformation and content or performance drift (eg, model updates, guideline changes, prompt template changes, and channel shifts) pose direct hazards to patient decision-making [15]. Hyperrealistic “clinician” avatars can convey inaccurate advice with a credibility premium that textual chatbots rarely command. Subtle script hallucinations and the lack of standardized clinical review workflows in many deployments amplify the chance that viewers will act on falsehoods before clinicians can intervene [16,17]. These risks are heightened in asynchronous, public-facing channels where corrections lag behind dissemination and platform ranking may preferentially surface engaging content.

Second, identity misuse and privacy infringements are uniquely salient when the video is the medium because identity cues drive trust calibration and downstream adherence. Unauthorized cloning of a clinician or patient’s likeness undermines autonomy, consent, and informational self-determination. Even ostensibly therapeutic recreations such as those involving deceased relatives raise unresolved questions about posthumous privacy and family interests [18,19]. Because mere visual plausibility confers trust, impersonation can catalyze fraud and degrade the informational environment far beyond the index case.

Third, psychological impact is bidirectional and context dependent. Video immersive qualities can deepen engagement, although they may also retraumatize, induce overattachment to synthetic figures, or blur boundaries between memory and simulation in grief and trauma work [20,21]. Minimizing these harms requires careful screening, clear framing, and predefined discontinuation criteria with escalation pathways to human care—not only technical guardrails.

Operationalizing these concerns should prioritize minimal, measurable guardrails. For comprehension and misinformation, institutions should track user-reported confusion, unplanned follow-up contacts attributable to the content, and brief comprehension checks (eg, teach back–style questions) in representative patient groups [22]. For identity and autonomy, all patient-facing deployments should meet a baseline disclosure standard, including clear labeling of synthetic content, explicit affirmation that no real clinician is speaking, and an accessible opt out. Psychological risk warrants prescreening, short validated distress scales, and predefined stop rules. Equity should be audited through stratified analyses of comprehension, incident or complaint rates, and follow-up burden (eg, by language, age, and health literacy). These metrics make benefits and harms visible enough to guide iterative redesign and trigger rereviews when thresholds are crossed.

Fourth, authenticity and institutional trust are collective goods at stake. As synthetic media saturate telehealth, patients may begin to doubt legitimate communications (“Is this my doctor or an AI?”) [23,24]. The resulting frictions—hesitation to follow instructions and demand for redundant confirmation—impose hidden costs on clinicians and organizations. Therefore, provenance signals and disclosure norms matter not as mere formalities but as trust-preserving infrastructure.

Finally, justice and equity considerations cut across all preceding risks. Benefits may accrue first to well-resourced settings that can build multilingual, culturally attuned avatars, whereas harms—deception, confusion, and exploitation—disproportionately fall on groups with lower health literacy or access to verification tools [25,26]. Thus, equity-oriented design, performance disaggregation, and complaint path accessibility are ethical requirements, not optional enhancements.


We present an operational governance framework—the risk-and-ethics matrix—that supports deployment decisions for patient-facing AI-generated video by linking residual clinical risk to a principlism-based EAS and to predefined monitoring and rereview actions. Existing laws, policies, and AI governance frameworks establish essential high-level expectations. They do not usually specify how institutions should adjudicate a concrete patient-facing AI-generated video use case, what minimum controls should accompany approval, or when iterative changes should trigger rereview. Purely technical risk scoring tends to underweight autonomy and justice [27,28], whereas principle-first approaches can ignore how likely and severe harms are in actual practice [28-31]. Our integration preserves the strengths of both approaches and translates them into decisions that IRBs, hospital digital health and patient education governance groups, communications leaders, and safety and IT oversight committees can defend, document, and audit.

Here, we distinguish inherent or unmitigated clinical hazard from residual clinical risk, which is the basis for governance classification. On the risk axis, we adapt a hospital-grade matrix consistent with common clinical risk management concepts in which risk reflects the combination of probability and severity. For each use case, we score (1) the likelihood that a specified harm scenario will occur on a 4-level ordinal scale (rare, unlikely, possible, and likely) and (2) the severity of plausible consequences on a 4-level scale (negligible, minor, major, and catastrophic). Cross-tabulation yields composite tiers of low, moderate, high, and extreme residual risk. Assessment proceeds by enumerating the failure modes specific to synthetic video—misinformation that could change care, identity or privacy breaches, psychologically triggering content, or equity harms [19,32,33]—and then rating each mode and assigning the overall tier according to the highest credible residual risk after proposed mitigations. Mitigations—such as clinician review of scripts, constrained generation, authenticated distribution, and provenance or disclosure controls—are recorded in the dossier with versioning and change logs so that residual (not theoretical) risk is the basis of classification over time.

On the ethics side, we operationalize the 4 principles—autonomy, beneficence, nonmaleficence, and justice—into an EAS ranging from 0 to 8. Each principle receives a score of 0 when violated, a score of 1 when partially upheld, and a score of 2 when clearly upheld, guided by concrete criteria that map abstract duties to observable practices. Autonomy considers the transparency of AI use, the accuracy of identity representation, voluntariness, and the adequacy of consent in a video medium [34]; beneficence requires a credible, evidence-informed benefit that is proportionate to the foreseeable burdens [35]; nonmaleficence emphasizes minimizing physical, psychological, informational, and reputational harms and guarding against foreseeable misuse [36]; and justice attends to equitable access and performance across groups, bias mitigation, nonexploitation of vulnerable populations, and preservation of public trust [37]. We band the EAS scores as high (7-8), medium (4-6), or low (0-3); where evidence is limited, conservative scoring and explicit uncertainty statements are needed. Table 1 shows the scales and rubric.

Table 1. Scales and rubric for the risk-and-ethics matrix.
Component and level or principleDefinition or criterion
Likelihood
RareRare under routine conditions; requires multiple safeguards to fail
UnlikelySingle lapse or unusual context
PossibleCommon precursor conditions present
LikelyReproducible under routine conditions
Severity
NegligibleNo decision impact; self-correcting (eg, brief uncertainty without behavior change)
MinorTransient confusion; extra contact (eg, 1 follow-up call or portal message for clarification)
MajorClinically consequential misinformation or marked psychological harm (eg, delayed care, inappropriate self-management, or significant distress requiring clinician intervention)
CatastrophicSevere harm or system-level misinformation (eg, serious injury, widespread harmful misinformation, or crisis-level psychological destabilization)
EASa—autonomy
0No or unclear disclosure; misleading identity; no opt out
1Disclosure present but incomplete or hard to understand
2Clear disclosure; accurate identity; voluntary, informed consent
EAS—beneficence
0No credible benefit
1Plausible benefit; limited evidence
2Evidence-informed benefit; proportional to burdens
EAS—nonmaleficence
0Foreseeable significant harms
1Harms possible with partial mitigation
2Robust mitigation (human in the loop, constrained generation, or crisis plan)
EAS—justice
0Exacerbates inequity or exploitation
1Neutral or unclear
2Equitable access; bias mitigation; accessible complaint path
EAS banding
High7-8
Medium4-6
Low0-3

aEAS: ethical alignment score.

Residual clinical risk is scored on a 4-level likelihood scale (rare, unlikely, possible, and likely) and a 4-level severity scale (negligible, minor, major, and catastrophic) interpreted after proposed mitigations. Ethical alignment is scored using the EAS (0-8), operationalizing autonomy, beneficence, nonmaleficence, and justice on a scale from 0 to 2 per principle and then banded as high (7-8), medium (4-6), or low (0-3). Ratings should reflect residual (not theoretical) risk, with conservative defaults and explicitly recorded uncertainty when evidence is sparse, and should be versioned for rereview after material changes.

Plotting the residual risk tier against EAS bands yields 4 deployment dispositions with explicit entry rules (Figure 1): encourage, permit with oversight, restrict or redesign, and prohibit. “Encourage” applies when residual risk is low and ethical alignment is high, supporting routine deployment with disclosure and periodic quality assurance. “Permit with oversight” covers moderate risk with at least medium ethical alignment (or low risk with medium ethical alignment) and requires human-in-the-loop review where appropriate, audit trails, incident reporting, time-limited approvals, and a postdeployment monitoring plan with predefined triggers for rereview. “Restrict or redesign” is appropriate when residual risk is high or when ethical alignment is low in the absence of intrinsic deception (ie, the use case is not fundamentally based on impersonation, undisclosed synthetic clinicians, or manipulative identity cues); here, scope narrowing, stronger transparency and safety guardrails, and protocolized pilots are prerequisites for reconsideration. “Prohibit” is reserved for extreme risk or for low ethical alignment tied to intrinsic deception or manipulation that exploits vulnerabilities; in such cases, deployment is disallowed and takedown, reporting, and other platform- or legal-level remedies may be warranted. These thresholds aim less for numerical precision than for defensible consistency across cases and clarity about what concrete changes would move an application toward a safer, more ethically aligned quadrant.

Figure 1. Risk-and-ethics matrix for patient-facing generative artificial intelligence video in health care. Residual clinical risk (likelihood × severity after mitigations) is plotted against ethical alignment (ethical alignment score; EAS), yielding 4 deployment dispositions: encourage, permit with oversight, restrict or redesign, and prohibit. The horizontal axis represents the residual risk tier (low, moderate, high, or extreme), and the vertical axis represents EAS band (high, medium, and low), where EAS operationalizes autonomy, beneficence, nonmaleficence, and justice on a scale from 0 to 8. Entry rules prioritize residual (not theoretical) risk and link each disposition to minimum controls and rereview triggers.

To support consistency, the rubric anchors abstract principles to concrete artifacts (eg, disclosure language, escalation pathways, evidence of benefit, and provenance controls) so that different panels can converge even when data are sparse [38,39]. Interrater reliability is promoted through independent prescoring, structured reconciliation, and written rationales for deviations from precedent. Because both risk and ethics are provisional in fast-moving sociotechnical contexts [40], institutions should version scores with date-stamped assumptions and require rereview after model updates, distribution channel changes, or sentinel events. Therefore, classification is not a verdict but a living record of judgment under stated conditions.

Finally, we specify a lightweight workflow that fits existing governance rather than creating parallel structures. Because review burden should be proportional to risk and novelty, we do not assume a single fixed evaluation time for all use cases. Low-risk, template-based, clinician-vetted educational videos that closely follow prior approved formats may undergo an expedited review focused on dossier updates, disclosure, and any material changes, whereas novel, higher-risk, psychologically sensitive, identity-based, or publicly disseminated use cases warrant fuller panel deliberation and documentation. Proponents submit a use case dossier describing purpose and audience, generation pipeline, distribution channel, mitigation plan, anticipated failure modes, and versioning or change logs. A triad panel—a clinician or health educator, bioethicist, and safety or IT lead—scores risk and EAS independently, reconciles differences, and documents residual disagreements; panels may co-opt patient advocacy or health literacy expertise for patient-facing deployments when needed. Decisions link directly to the chosen disposition and to a postdeployment monitoring plan with predefined indicators (eg, misinformation incidents, user-reported confusion, complaint rates, follow-up burden, and equity gaps) and time-bound rereview triggers. Operational steps are summarized in Figure 2; a printable evaluator’s checklist and monitoring triggers can be found in Multimedia Appendix 1, and a structured use case dossier template can be found in Multimedia Appendix 2. For institutional use, matrix-guided evaluation should be embedded into release governance such that publication or distribution through official channels requires completed dossier documentation, named sign-off, and versioned approval records. Materials that bypass review or breach minimum controls should trigger withholding of institutional dissemination; pause or takedown; incident review; and, where applicable, corrective action under local policy. In this sense, the key incentive structure is not speed-based reward. It is the linkage of authorization, traceability, and consequences to the right to deploy patient-facing AI-generated video. Where feasible, dossiers should trace content provenance from prompt to rendered asset to distribution and specify abuse-resistant defaults (eg, prohibited impersonation classes and persistent labeling), with escalation procedures when out-of-distribution use or unequal performance emerges. Table 2 shows governance categories and minimum controls.

Figure 2. Workflow from use case dossier to classification and postdeployment monitoring. Proponents submit a versioned use case dossier; a triad panel (clinician or health educator, bioethicist, and safety or IT lead) independently prescores residual risk and ethical alignment score (EAS), reconciles differences, and records a written rationale. The resulting disposition maps to minimum controls (Table 2) and a monitoring plan with predefined indicators and rereview triggers (eg, incidents, user-reported confusion, equity gaps, and version changes). Operational checklist and monitoring triggers are provided in Multimedia Appendix 1; the dossier template is provided in Multimedia Appendix 2.
Table 2. Deployment dispositions, entry rules, and minimum controls.
CategoryResidual risk tier × EASa bandMinimum controlsbIllustrative examplesc
EncourageLow risk and high EASDisclosure and routine QAdTransparent, clinician-vetted patient education avatar
Permit with oversightModerate risk and ≥medium EAS or low risk and medium EASHuman in the loop, audit trail, incident reporting, and time-limited approvalProtocolized “deepfake therapy” in specialist or IRBe settings
Restrict or redesignHigh risk or low EAS (no intrinsic deception)Scope narrowing, stronger transparency and safety, and piloting then rescoringFree-text explainer without clear disclosure
ProhibitExtreme risk or low EAS with intrinsic deception or manipulationTakedown and platform- or legal-level remediesImpersonated physician endorsements

aEAS: ethical alignment score.

bMinimum controls specify documentation, human oversight, incident reporting, time-limited approval, and rereview triggers as applicable.

cExamples illustrate typical placements.

dQA: quality assurance.

eIRB: institutional review board.


The following use cases illustrate a repeatable mapping from dossier evidence to residual risk or EAS scoring, a deployment disposition, and a minimal monitoring plan with explicit triggers for rereview.

Case 1: Patient Education Avatar (Transparent, Vetted, and Multilingual)

A hospital deploys short, tailored videos explaining procedures and postoperative care through a clearly disclosed AI avatar delivered via an authenticated patient portal. Scripts are clinician vetted, linguistically and culturally adapted, and version controlled. The portal supports replay, adjustable playback speed, and an easy pathway to request human follow-up or report confusion. Residual risk is low: under constrained scripts plus clinical review and secure distribution, misinformation is rare or unlikely and typically minor (eg, transient confusion or extra contact), and privacy exposure is limited because the content is generic. Ethical alignment would likely be high (autonomy=2, beneficence=2, nonmaleficence=2, and justice=2): disclosure and an opt out support autonomy, measurable gains in comprehension and reduced avoidable follow-up burden support beneficence, constrained generation and review reduce foreseeable harms, and multilingual access advances justice. The resulting disposition is to encourage, with routine quality assurance and a time-bounded review cadence plus rereview triggers for guideline changes, prompt template updates, channel changes, or stratified performance gaps across language and health literacy groups [41,42].

Case 2: “Deepfake Therapy” for Grief (Therapist Led, Protocolized, and Time Limited)

Under IRB-approved protocols, a psychotherapist offers a time-bounded intervention in which a synthetic likeness of a deceased relative delivers scripted messages to facilitate goodbye rituals. A research or specialist setting matters because it enables structured screening, standardized outcome capture, adverse event reporting, and enforceable stop rules. Residual risk is moderate to high: even with careful preparation, clinically meaningful psychological harms may occur (severity level: major), and the likelihood may be possible or likely in vulnerable subgroups. Ethical alignment would likely be medium (autonomy=1, beneficence=2, nonmaleficence=1, and justice=1): beneficence may be credible for selected patients; autonomy depends on robust consent that frames the video as a simulation and checks understanding; nonmaleficence relies on screening, therapist presence, predefined discontinuation criteria, and escalation pathways; and justice requires equitable access criteria and avoidance of coercive commercialization. The resulting disposition is to permit with oversight only in research or specialist settings, with predefined outcome measures (including follow-up windows), incident logging, and immediate cessation upon adverse reactions; approvals should be time limited, with rereview triggers tied to protocol deviations, adverse events, or material changes to the model or pipeline [5,43,44].

Case 3: Impersonated Physician Endorsements (No Consent and Public Dissemination)

A synthetic clone of a prominent clinician appears on social media platforms to promote unverified health products or claims. Residual risk is extreme: harm is likely because the video exploits identity-based trust and can divert patients from evidence-based care; severity level is major to catastrophic if it prompts medication changes, delays appropriate care, or amplifies misinformation at scale. Ethical alignment would score low across all 4 principles: deception negates autonomy; benefits are not patient centered; harms are foreseeable and unmitigated; and the practice exploits vulnerable audiences, undermining justice and public trust. The resulting disposition is to prohibit. Rapid takedown and reporting workflows, provenance checks, public clarification through verified institutional channels, and legal remedies are warranted; organizations should precommit to a zero-tolerance policy for unauthorized likeness use and log incidents to strengthen postmarket monitoring and prevention [44-46].


Implications for Digital Health Implementation

Our risk-and-ethics matrix translates life cycle AI governance expectations into a practical format for clinical decision-makers by pairing probabilistic risk appraisal with principled ethics in a way that is explicit, documentable, and revisitable. The distinctive governance value of the risk-and-ethics matrix is that it bridges the gap between high-level regulatory expectations and local operational decisions. Rather than offering another abstract set of principles, it enables institutions to classify a specific use case; document the rationale for approval or restriction; assign minimum controls; and connect deployment to monitoring, incident response, and change control over time. This orientation is consistent with major governance frameworks that emphasize postdeployment monitoring, mechanisms for capturing user input, incident response, and change management as core components of responsible AI use in real-world settings [47]. We localize these expectations to patient-facing generative AI video by grounding “risk tiering” in concrete clinical harm scenarios (eg, misinformation that changes care, identity misuse, psychological triggering content, and equity harms) and by converting principlism—autonomy, beneficence, nonmaleficence, and justice—into action-guiding criteria through an EAS. Together, the 2 axes yield implementable dispositions—encourage, permit with oversight, restrict or redesign, and prohibit—whose entry rules can be recorded, audited, and defended as part of evaluating real-world clinical performance.

Beyond classification, the matrix provides a shared structure for interdisciplinary deliberation and practical redesign. The framework also functions as a design instrument: because dossiers trace why a proposal lands in a given disposition, developers are directed toward concrete modifications—clear disclosure and comprehension-checked consent to strengthen autonomy; scope limitation, constrained generation, and human-in-the-loop review to reduce residual risk; and stratified monitoring to strengthen justice. In parallel, professional guidance on generative AI in medicine underscores the importance of preserving human oversight and aligning deployments with clinical workflows rather than displacing them—an emphasis that is especially salient for persuasive patient-facing media. Conversely, the matrix clarifies “red-line” cases grounded in intrinsic deception (eg, impersonated clinician endorsements), supporting rapid takedown, institutional clarification through verified channels, and incident logging to strengthen future prevention.

National-level safeguards become especially important in cases in which institutional incentives favor speed, visibility, or monetization over careful review. In China, the relevant governance architecture is emerging but remains distributed across health sector, platform, and generative AI rules. Existing measures already provide building blocks, including synthetic content labeling and traceability, filing and disclosure for certain public-facing AI services, and health sector expectations for account registration and monitoring. The next step is to connect these elements through sector-specific requirements for disclosure, provenance, verified identity, monitoring, rapid correction or takedown, and enforceable accountability. Such national-level guardrails would not replace local review tools such as the risk-and-ethics matrix; rather, they would create the incentive environment in which institutional review is more likely to be performed seriously and consistently.

Feasibility in smaller or resource-limited settings will depend on tiered implementation rather than assuming the full model from the outset. The core minimum is not a large committee but a documented review pathway with clearly assigned accountability, use case documentation, explicit disclosure checks, and a mechanism for escalation when risk exceeds local expertise. For familiar low-risk educational videos, institutions with limited resources may use a simplified pathway involving a clinically accountable reviewer plus a second reviewer with operational or technical oversight supported by a standardized checklist and basic postdeployment signals such as complaints, follow-up contacts, and disclosure failures. The fuller triad panel model—with dedicated bioethics input, richer analytics, formal incident reporting, and equity stratification—should be viewed as an expanded configuration for higher-risk or more mature settings. Where dedicated bioethics expertise is unavailable, regional ethics consortia, shared review pools, tele-ethics consultation, or referral pathways to larger centers may provide a practical alternative, especially for first-in-class, psychologically sensitive, identity-based, or publicly disseminated use cases.

Limitations and Validation Agenda

This approach has limitations. Early deployments will often rely on expert judgment because empirical evidence on the frequency and magnitude of novel harms remains sparse, and both residual risk estimates and EAS components can vary with local context [48-51]. To temper subjectivity, we emphasize independent prescoring, structured reconciliation, and written rationales anchored to observable artifacts (eg, disclosure language, evidence of benefit, escalation pathways, and provenance controls) [52]. Because the EAS is intended as an operational rubric rather than a purely intuitive checklist, institutions should prospectively calibrate and evaluate its reliability before routine use. A practical approach would be to begin with a set of anchor case vignettes spanning low-, medium-, and high-alignment scenarios; require independent prescoring by panel members; conduct structured reconciliation with written reasons for disagreement; and repeat calibration periodically using shared cases across panels or sites [53]. For the ratings of 0 to 2 assigned to each principle, agreement could be summarized using percentage agreement and weighted κ, whereas the reliability of the summed EAS from 0 to 8 could be examined using an intraclass correlation coefficient. Institutions could additionally track agreement on EAS banding and on the final deployment disposition because these outputs are directly tied to governance decisions. Content validity could be strengthened through multidisciplinary expert review of whether the rubric adequately captures observable manifestations of autonomy, beneficence, nonmaleficence, and justice, with iterative refinement through pilot-testing or Delphi-style consensus procedures [54]. Construct validity could then be explored by testing whether the EAS discriminates between use cases that are expected a priori to differ in ethical alignment (eg, transparent clinician-vetted education avatars vs impersonated clinician endorsements) [55]. Importantly, classification should be treated as a living record—versioned with date-stamped assumptions—so that uncertainty becomes auditable and revisable rather than implicit [56,57]. The same use case may plausibly yield different profiles across clinical domains (eg, perioperative education vs mental health) or across resource settings; documenting contextual assumptions and applying prespecified domain modifiers can improve consistency without suppressing legitimate local variation. Finally, risk severity overlaps conceptually with nonmaleficence, and beneficence often embeds risk-benefit trade-offs [58,59]. Therefore, treating the axes as orthogonal is a usability heuristic—intended to promote clarity and reproducibility—rather than a claim of theoretical independence; cross-referencing during deliberation should be expected.

Measurement, Monitoring, and Rereview Triggers

These caveats point to a concrete real-world evaluation agenda. Retrospective incident reviews and prospective pilots can stress test thresholds and calibrate rubrics using patient-centered and workflow-relevant end points (eg, comprehension or teach back performance, unplanned follow-up contacts attributable to content, complaint and incident rates, and stratified equity gaps). Textbox 1 summarizes a core monitoring set, operational data sources, and example rereview triggers that can be embedded into routine digital health workflows. Governance-relevant measurement should be paired with life cycle mechanisms for capturing user input and adjudicating overrides and with explicit incident response and recovery pathways—elements foregrounded in the National Institute of Standards and Technology’s risk management guidance for deployed AI systems [14]. Harmonization with provenance and disclosure standards can further improve auditability and reduce identity-related misuse, enabling versioned reassessments as models, prompts, distribution channels, and guardrails evolve. For systems that will undergo iterative change, “change control” should be planned rather than improvised; regulatory thinking around predetermined change control plans provides a useful template for specifying anticipated modifications and the evidence required to validate them over time. Comparative ethical analysis may also be valuable in edge cases where principlism and alternative lenses diverge; documenting such divergences can refine decision rules while preserving usability [60,61].

Textbox 1. Core monitoring metrics, operational data sources, and rereview triggers for patient-facing artificial intelligence–generated video.

Core metrics (minimum set)

  • Comprehension: brief teach back–style checks or short postview questions and user-reported confusion
  • Content-attributable follow-up burden: messages, calls, or telehealth follow-ups attributable to the video (eg, tagged reason codes or postview “contact clinician” clicks)
  • Incidents and complaints: safety reports and formal complaints linked to the content (misinformation, identity misuse, or privacy concerns)
  • Equity gaps: stratified differences in comprehension, follow-up burden, and incidents or complaints (eg, by language and health literacy proxies)
  • Provenance and trust: visibility of synthetic content disclosure, verification friction (eg, “Is this my doctor?” queries), and confirmed impersonation attempts

Data sources (digital health operations)

  • Patient portal and telehealth analytics (views, completion, and click-through to contact or obtain support)
  • Secure messaging and call center logs (tags and reason codes and clinician note templates for attribution)
  • Incident reporting and complaint systems (patient safety and privacy or security tickets)
  • Patient feedback channels (postview survey and “report confusion/request human help” buttons)
  • If distributed externally, verified channel monitoring (eg, takedown requests and platform reports)

Rereview triggers (examples; adapt locally)

  • Material change: model, prompt or template, script or guideline, channel, or language rollout updates
  • Signal excursion: sustained rise in follow-up burden or confusion reports vs baseline
  • Safety event: any major incident or clustered minor incidents attributable to the content
  • Equity flag: new or widening stratified gaps in end points
  • Provenance or identity event: confirmed impersonation, unauthorized likeness use, or disclosure failure

In early deployments, trigger thresholds should be treated as provisional rather than fixed regulatory cutoff points. A pragmatic starting approach is to establish a local baseline during an initial pilot or first complete rollout cycle and then update thresholds iteratively as more observations accumulate. During this early phase, institutions may use structured expert consensus or Delphi-style calibration among early adopters to define provisional trigger ranges, with wider tolerance bands and explicit uncertainty notes until local rates stabilize. Thresholds should ideally be interpreted against rolling local baselines rather than in isolation and should be re-estimated after material workflow, model, channel, or language changes. For equity analyses in particular, subgroup differences should not be overinterpreted when denominators are sparse; institutions should prespecify minimum subgroup counts before treating observed gaps as decision relevant and, where counts remain small, use descriptive flagging and continued data collection rather than strong statistical conclusions.

Here, “rereview” denotes the procedural reassessment triggered by monitoring signals, incidents, or material changes. “Reclassification” denotes a substantive change in risk tier, EAS band, or deployment disposition that may result from that reassessment. Because ethical priorities and risk tolerance vary across cultures and health systems, the framework is designed to be portable yet tunable [62,63]. To preserve comparability while allowing for local adaptation, the framework distinguishes core elements that should remain stable across sites from parameters that may be tuned to local context (Table 3).

Table 3. Core standardized elements and locally tunable parameters of the risk-and-ethics matrix.
DomainCore elements (preserve across sites)aTunable parameters (adapt locally)b
Ethical architectureFour principles; 0-2 EASc scoring logic; high, medium, and low EAS bandingLocal case anchors, examples, and training materials
Residual risk assessmentResidual risk logic after mitigation; 4-level likelihood and severity structureDefault risk modifiers for high-vulnerability domains or populations
Deployment decisionsFour disposition categories (encourage, permit with oversight, restrict or redesign, and prohibit)Local approving body, escalation route, and implementation authority
Documentation and oversightUse case dossier, rationale, versioning, accountability, and human oversightDossier format; full triad panel vs simplified or shared review pathway
Disclosure and provenanceMinimum synthetic content disclosure, identity accuracy, and provenance expectationsDisclosure wording, language level, format, and content credential implementation
Monitoring and rereviewMonitoring, incident capture, equity review, and rereview trigger logicIndicator thresholds, baseline methods, observation windows, subgroup definitions, and language coverage minimums

aCore elements are intended to preserve conceptual and operational comparability across sites.

bTunable parameters may be adapted to local workflow capacity, legal context, language needs, and risk tolerance provided that deviations are specified prospectively and documented consistently within the adopting institution.

cEAS: ethical alignment score.

Crucially, classification must remain revisitable. Approvals in “permit with oversight” should be time limited and coupled with predefined rereview triggers (eg, model or prompt updates, distribution channel changes, sentinel incidents, or emerging inequities). Successful mitigation and accumulating evidence may move a use case toward “encourage,” whereas incidents or drift may push it toward “restrict or redesign” or “prohibit.” Embedding this cadence operationalizes continuous risk management and aligns oversight with the broader shift toward postdeployment monitoring systems for AI in real-world settings.

The risk-and-ethics matrix is intended to complement, not replace, formal regulatory review. Its role is to translate broad regulatory expectations into case-level institutional governance for patient-facing AI-generated video use. In cross-jurisdictional terms, the framework’s disclosure, identity, and provenance elements align with transparency-oriented requirements; its human-in-the-loop review, escalation pathways, and authority to pause or withdraw deployments align with expectations around human oversight; its dossier, versioning, and documented rationale align with technical documentation and record-keeping expectations; and its monitoring indicators, incident triggers, and rereview cadence align with postmarket monitoring and iterative change control requirements. This means that, where a use case is already subject to sector-specific regulation—such as the European Union AI Act’s risk-based obligations for certain AI systems or medical device review pathways that incorporate predetermined change control planning—the matrix is not a substitute for those legal processes. Rather, it provides a local operational layer that helps institutions implement, document, and monitor responsible use under real-world conditions.

Conclusions and Next Steps for Implementation

AI-generated video is becoming a routine modality for patient-facing communication, making life cycle governance essential to safe real-world deployment [1,64,65]. This Viewpoint presents an operational risk-and-ethics matrix that links residual clinical risk and ethical alignment to auditable, revisitable deployment decisions.

Future work should prioritize 3 deliverables. First, real-world evaluation should calibrate thresholds with patient-centered and workflow-relevant end points (eg, comprehension and teach back, content-attributable follow-up burden, incidents and complaints, and stratified equity gaps) and embed monitoring, incident response, and change management with time-limited approvals and rereview triggers. Second, provenance should be strengthened through clear disclosure and interoperable content credentials to reduce identity misuse and support verification. Third, change control should be planned rather than improvised: institutions should prespecify anticipated model, prompt, or pipeline updates and the evidence required to maintain assurance over time, drawing on the Food and Drug Administration’s predetermined change control plan approach for AI-enabled systems.

Acknowledgments

The authors used ChatGPT (OpenAI) for limited language translation and language editing during manuscript preparation. The authors reviewed and revised all such output and take full responsibility for the final manuscript.

Funding

This research was supported by grants from the National Natural Science Foundation of China (82370724), the Qingdao Key Health Discipline Development Fund, and the Qingdao Key Clinical Specialty Elite Discipline project.

Authors' Contributions

WJ and YH conceived the presented idea and developed the theoretical framework for the risk-and-ethics evaluation scaffold for patient-facing generative artificial intelligence video. YH conducted the literature synthesis, drafted the manuscript, and designed the figures. WJ supervised the project, secured funding, and provided critical revision of the manuscript for important intellectual content. Both authors discussed the concepts and approved the final version of the manuscript.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Operational tool pack (evaluator’s checklist and monitoring triggers). This pack operationalizes the risk-ethics matrix for routine governance. Part A provides a 10- to 12-item evaluator’s checklist. Part B lists monitoring indicators with example triggers. Crossing a trigger prompts rereview and reclassification per Table 2 and Figure 2.

DOC File, 314 KB

Multimedia Appendix 2

Use case dossier template (versioned submission form). The dossier captures purpose and audience, generation pipeline, distribution channel, content and privacy, failure modes with mitigations, ethical alignment score (EAS) evidence (0-2 per principle), residual risk tier, governance ask, monitoring plan and thresholds, sign-offs, and change log. It aligns with Table 1 (scales and EAS rubric), Figure 1 (quadrant mapping), Table 2 (entry rules and minimum controls), and Figure 2 (workflow). Panels perform independent prescores, reconciliation, and written rationales; approvals in yellow zones are time limited, with predefined triggers for rereview.

DOC File, 236 KB

  1. Netland T, von Dzengelevski O, Tesch K, Kwasnitschka D. Comparing human-made and AI-generated teaching videos: an experimental study on learning effects. Comput Educ. Jan 2025;224:105164. [CrossRef]
  2. Lin J, Gu Y, Du G, et al. 2D/3D image morphing technology from traditional to modern: a survey. Inf Fusion. May 2025;117:102913. [CrossRef]
  3. Rovati L, Gary PJ, Cubro E, et al. Development and usability testing of a patient digital twin for critical care education: a mixed methods study. Front Med (Lausanne). 2024;10:1336897. [CrossRef] [Medline]
  4. Islam MZ, Wang G. Avatars in the educational metaverse. Vis Comput Ind Biomed Art. Jun 10, 2025;8(1):15. [CrossRef] [Medline]
  5. Hoek S, Metselaar S, Ploem C, Bak M. Promising for patients or deeply disturbing? The ethical and legal aspects of deepfake therapy. J Med Ethics. Jul 7, 2025;51(7):481-486. [CrossRef] [Medline]
  6. Düzgün MV, İşler A, Kazan MS. Effect of avatar-based education program in hydrocephalus on ventriculoperitoneal shunt complications and parents’ knowledge and care skills: multicenter randomized controlled trials. Pediatr Neurol. Aug 2025;169:131-139. [CrossRef] [Medline]
  7. Queiroz ABL, Sartori LRM, Lima GDS, Moraes RR, Lima GS. Editorial policies for use and acknowledgment of artificial intelligence in dental journals. J Dent. Oct 2025;161:105923. [CrossRef] [Medline]
  8. Wekenborg MK, Gilbert S, Kather JN. Examining human-AI interaction in real-world healthcare beyond the laboratory. NPJ Digit Med. Mar 19, 2025;8(1):169. [CrossRef] [Medline]
  9. Eutamene HB, Hamidouche W, Keita M, Taleb-Ahmed A, Hadid A. Integrating perceptual quality analysis and caption-based features for robust deepfake video detection. Comput Electr Eng. Dec 2025;128:110699. [CrossRef]
  10. Benezeth Y, Krishnamoorthy D, Botina Monsalve DJ, Nakamura K, Gomez R, Mitéran J. Video-based heart rate estimation from challenging scenarios using synthetic video generation. Biomed Signal Process Control. Oct 2024;96:106598. [CrossRef]
  11. Zahedi FM, Zhao H, Sanvanson P, Walia N, Jain H, Shaker R. My real avatar has a doctor appointment in the wepital: a system for persistent, efficient, and ubiquitous medical care. Inf Manag. Dec 2022;59(8):103706. [CrossRef]
  12. Sestino A, D’Angelo A. My doctor is an avatar! The effect of anthropomorphism and emotional receptivity on individuals’ intention to use digital-based healthcare services. Technol Forecast Soc Change. Jun 2023;191:122505. [CrossRef]
  13. Decety J, Li J. The value of empathy in medical practice: a neurobehavioral perspective. Soc Sci Humanit Open. 2025;12:101956. [CrossRef]
  14. Tabassi E. Artificial intelligence risk management framework (AI RMF 1.0). National Institute of Standards and Technology; 2023. URL: https://www.nist.gov/publications/artificial-intelligence-risk-management-framework-ai-rmf-10 [Accessed 2026-04-27]
  15. Masayoshi K, Katada Y, Ozawa N, Ibuki M, Negishi K, Kurihara T. Deep learning segmentation of non-perfusion area from color fundus images and AI-generated fluorescein angiography. Sci Rep. May 11, 2024;14(1):10801. [CrossRef] [Medline]
  16. Piot MA, Attoe C, Billon G, Cross S, Rethans JJ, Falissard B. Simulation training in psychiatry for medical education: a review. Front Psychiatry. 2021;12:658967. [CrossRef] [Medline]
  17. Zhang J, Ding Y, Zhang H, et al. An experimental study on embodiment forms and interaction modes in affective robots for anxiety relief and emotional connection. Int J Hum Comput Stud. Sep 2025;203:103584. [CrossRef]
  18. Soto-Sanfiel MT, Wu Q. How audiences make sense of deepfake resurrections: a multilevel analysis of realism, ethics, and cultural meaning. Comput Hum Behav. Jan 2026;174:108822. [CrossRef]
  19. Ma’arif A, Maghfiroh H, et al. Social, legal, and ethical implications of AI-generated deepfake pornography on digital platforms: a systematic literature review. Soc Sci Humanit Open. 2025;12:101882. [CrossRef]
  20. Wagner R, Pardi G, Müller J, Brucker B, Schwarzer S, Gerjets P. Listening to scientists in immersive videos: how levels of immersion and points of view influence learning experiences. Comput Educ. 2025;234:1-19. [CrossRef]
  21. Bujić M, Salminen M, Hamari J. Effects of immersive media on emotion and memory: an experiment comparing article, 360-video, and virtual reality. Int J Hum Comput Stud. Nov 2023;179:103118. [CrossRef]
  22. Fridman I, Bylund CL, Elston Lafata J. Trust of social media content and risk of making misinformed decisions: survey of people affected by cancer and their caregivers. PEC Innov. 2024;5:100332. [CrossRef] [Medline]
  23. Diwanji VS. Should your brand hire virtual influencers? How realism and gender presentation shape trust and purchase intentions. J Retail Consum Serv. Jan 2026;88:104491. [CrossRef]
  24. Letafati M, Otoum S. On the privacy and security for e-health services in the metaverse: an overview. Ad Hoc Netw. Nov 2023;150:103262. [CrossRef]
  25. Pirhofer J, Bükki J, Vaismoradi M, Glarcher M, Paal P. A qualitative exploration of cultural safety in nursing from the perspectives of Advanced Practice Nurses: meaning, barriers, and prospects. BMC Nurs. Jul 4, 2022;21(1):178. [CrossRef] [Medline]
  26. Sterponi L, Fatigante M, Zucchermaglio C, Alby F. Companions in immigrant oncology visits: uncovering social dynamics through the lens of Goffman’s footing and Conversation Analysis. SSM Qual Res Health. Jun 2024;5:100432. [CrossRef]
  27. Krijger J. What about justice and power imbalances? A relational approach to ethical risk assessments for AI. DISO. 2024;3:56. [CrossRef]
  28. Ploug T, Jørgensen RF, Motzfeldt HM, Ploug N, Holm S. The need for patient rights in AI-driven healthcare - risk-based regulation is not enough. J R Soc Med. Aug 2025;118(8):248-252. [CrossRef] [Medline]
  29. Baard P, Sandin P. Principlism and citizen science: the possibilities and limitations of principlism for guiding responsible citizen science conduct. Res Ethics. 2022;18(4):304-318. [CrossRef]
  30. Clouser KD, Gert B. A critique of principlism. J Med Philos. Apr 1990;15(2):219-236. [CrossRef] [Medline]
  31. Mittelstadt B. Principles alone cannot guarantee ethical AI. Nat Mach Intell. 2019;1(11):501-507. [CrossRef]
  32. Whittaker L, Kietzmann J, Letheren K, Mulcahy R, Russell-Bennett R. Brace yourself! Why managers should adopt a synthetic media incident response playbook in an age of falsity and synthetic media. Bus Horiz. 2023;66(2):277-290. [CrossRef]
  33. Mylrea M. The generative AI weapon of mass destruction: evolving disinformation threats, vulnerabilities, and mitigation frameworks. In: Lawless W, Mittu R, Sofge D, Fouad H, editors. Interdependent Human-Machine Teams: The Path to Autonomy. Academic Press; 2024:315-347. [CrossRef]
  34. Omrani N, Rivieccio G, Fiore U, Schiavone F, Agreda SG. To trust or not to trust? An assessment of trust in AI-based systems: concerns, ethics and contexts. Technol Forecast Soc Change. Aug 2022;181(2):121763. [CrossRef]
  35. Liao SM, Haykel I, Cheung K, Matalon T. Navigating the complexities of AI and digital governance: the 5W1H framework. J Responsible Technol. Sep 2025;23:100127. [CrossRef]
  36. Braga CM, Serrano MA, Fernández-Medina E. Towards a methodology for ethical artificial intelligence system development: a necessary trustworthiness taxonomy. Expert Syst Appl. Aug 2025;286:128034. [CrossRef]
  37. Pettersson M, Hedström M, Höglund AT. The ethics of DNR-decisions in oncology and hematology care: a qualitative study. BMC Med Ethics. Jul 31, 2020;21(1):66. [CrossRef] [Medline]
  38. Lanerolle G, Roberts ES, Haroon A, Shetty A. Introduction. In: Quality Assurance Management: A Comprehensive Overview of Real-World Applications for High Risk Specialties. Elsevier Science; 2024:1-21.
  39. Li S, Wang Z, Shang Y, et al. Developing federated time-to-event scores using heterogeneous real-world survival data. Comput Biol Med. Oct 2025;197(Pt B):111084. [CrossRef] [Medline]
  40. Patriarca R, Falegnami A, Costantino F, Di Gravio G, De Nicola A, Villani ML. WAx: an integrated conceptual framework for the analysis of cyber-socio-technical systems. Saf Sci. Apr 2021;136:105142. [CrossRef]
  41. Badawy MK, Khamwan K, Carrion D. A pilot study of generative AI video for patient communication in radiology and nuclear medicine. Health Technol. Mar 2025;15:395-404. [CrossRef]
  42. Adeboye W, Tayal V, Odubanjo E, et al. Artificial intelligence in the delivery of patient care: avatar-generated videos for patient education post breast surgery. Eur J Surg Oncol. May 2024;50:108076. [CrossRef]
  43. Emotional documentary explores new compassionate possibilities of VR. Unreal Engine. 2020. URL: https:/​/www.​unrealengine.com/​developer-interviews/​emotional-documentary-explores-new-compassionate-possibilities-of-vr?lang=fr [Accessed 2025-10-17]
  44. Pizzoli SF, Monzani D, Vergani L, Sanchini V, Mazzocco K. From virtual to real healing: a critical overview of the therapeutic use of virtual reality to cope with mourning. Curr Psychol. 2023;42(11):8697-8704. [CrossRef] [Medline]
  45. Liu T, Wang P, Pan D, Liu R. Credibility of AI generated and human video doctors and the relationship to social media use. Front Public Health. 2025;13:1559378. [CrossRef]
  46. Stokel-Walker C. Deepfakes and doctors: how people are being fooled by social media scams. BMJ. Jul 17, 2024;386:q1319. [CrossRef] [Medline]
  47. Nordling L. Scientists are falling victim to deepfake AI video scams - here’s how to fight back. Nature New Biol. Aug 7, 2024. [CrossRef] [Medline]
  48. Oehmen J, Locatelli G, Wied M, Willumsen P. Risk, uncertainty, ignorance and myopia: Their managerial implications for B2B firms. Industrial Marketing Management. Jul 2020;88:330-338. [CrossRef]
  49. Singh S, Dhumane A. Unmasking digital deceptions: an integrative review of deepfake detection, multimedia forensics, and cybersecurity challenges. MethodsX. 2025;15:103632. [CrossRef] [Medline]
  50. McAndrew T, Reich NG. An expert judgment model to predict early stages of the COVID-19 pandemic in the United States. PLoS Comput Biol. Sep 2022;18(9):e1010485. [CrossRef] [Medline]
  51. Awodi NJ, Liu YK, Ayodeji A, Adibeli JO. Expert judgement-based risk factor identification and analysis for an effective nuclear decommissioning risk assessment modeling. Prog Nucl Energy. Jun 2021;136:103733. [CrossRef]
  52. Maltby KM, Howes E, Lincoln S, et al. Marine climate change risks to biodiversity and society in the ROPME Sea Area. Clim Risk Manag. 2022;35(2121):100411. [CrossRef]
  53. Hallgren KA. Computing inter-rater reliability for observational data: an overview and tutorial. Tutor Quant Methods Psychol. 2012;8(1):23-34. [CrossRef] [Medline]
  54. Polit DF, Beck CT. The content validity index: are you sure you know what’s being reported? Critique and recommendations. Res Nurs Health. Oct 2006;29(5):489-497. [CrossRef] [Medline]
  55. Boateng GO, Neilands TB, Frongillo EA, Melgar-Quiñonez HR, Young SL. Best practices for developing and validating scales for health, social, and behavioral research: a primer. Front Public Health. 2018;6:149. [CrossRef] [Medline]
  56. Haltaufderheide J, Ranisch R. The ethics of ChatGPT in medicine and healthcare: a systematic review on Large Language Models (LLMs). NPJ Digit Med. Jul 8, 2024;7:183. [CrossRef] [Medline]
  57. Lekens AL, Drageset S, Hansen BS. Knowing how, arguing why: nurse anaesthetists’ experiences of nursing when caring for the surgical patient. BMC Nurs. Feb 7, 2025;24(1):144. [CrossRef] [Medline]
  58. Jansen SN, Kamphorst BA, Mulder BC, et al. Ethics of early detection of disease risk factors: a scoping review. BMC Med Ethics. Mar 5, 2024;25(1):25. [CrossRef] [Medline]
  59. Pathni RK. Beyond algorithms: ethical implications of AI in healthcare. Med J Armed Forces India. 2025;81(6):630-636. [CrossRef] [Medline]
  60. Rauprich O. Principlism. In: Chadwick R, editor. Encyclopedia of Applied Ethics. Academic Press; 2012:590-598. ISBN: 9780123739322
  61. Bello P, Bridewell W. Self-control on the path toward artificial moral agency. Cogn Syst Res. Jan 2025;89:101316. [CrossRef]
  62. Lysaght T, Chan HY, Scheibner J, Toh HJ, Richards B. An ethical code for collecting, using and transferring sensitive health data: outcomes of a modified Policy Delphi process in Singapore. BMC Med Ethics. Oct 4, 2023;24(1):78. [CrossRef] [Medline]
  63. Guenduez AA, Walker N, Demircioglu MA. Digital ethics: global trends and divergent paths. Gov Inf Q. Sep 2025;42(3):102050. [CrossRef]
  64. Vehi J, Mujahid O, Beneyto A, Contreras I. Generative artificial intelligence in diabetes healthcare. iScience. Aug 15, 2025;28(8):113051. [CrossRef] [Medline]
  65. Crowe B, Shah S, Teng D, et al. Recommendations for clinicians, technologists, and healthcare organizations on the use of generative artificial intelligence in medicine: a position statement from the Society of General Internal Medicine. J Gen Intern Med. Feb 2025;40(3):694-702. [CrossRef] [Medline]


AI: artificial intelligence
EAS: ethical alignment score
IRB: institutional review board


Edited by Amaryllis Mavragani; submitted 22.Jan.2026; peer-reviewed by Atefeh Shamsi, Zhaohui Su; final revised version received 14.Apr.2026; accepted 15.Apr.2026; published 08.May.2026.

Copyright

© Yongzheng Hu, Wei Jiang. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 8.May.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.